CMPS242 Final Project - A Comparison of Naive Bayes and Boosting
نویسنده
چکیده
My final project was to implement and compare a number of Naive Bayes and boosting algorithms. For this task I chose to implement two Naive Bayes algorithms that are able to make use of binary attributes, the multivariate Naive Bayes and the multinomial Naive Bayes with binary attributes. For the boosting side of the algorithms I chose to implement AdaBoost, and its close bother AdaBoost*. Both of these algorithms were chosen for their simplicity yet amazingly accurate results; in fact, the IEEE International Conference on Data Mining (ICDM) listed both AdaBoost and Naive Bayes as top ten algorithms in data mining [3].
منابع مشابه
A New Approach for Text Documents Classification with Invasive Weed Optimization and Naive Bayes Classifier
With the fast increase of the documents, using Text Document Classification (TDC) methods has become a crucial matter. This paper presented a hybrid model of Invasive Weed Optimization (IWO) and Naive Bayes (NB) classifier (IWO-NB) for Feature Selection (FS) in order to reduce the big size of features space in TDC. TDC includes different actions such as text processing, feature extraction, form...
متن کاملExperiments on Spam Detection with Boosting, Svm and Naive Bayes
For this project, I implement 3 popular text classification algorithms on spam detection, namely AdaBoost, Support Vector Machines and Naive Bayes. The performance are evaluated on some testing datasets. All experiments are done in Matlab. The experimental result is, all 3 algorithms have a satisfactory performance on spam detection. In term of accuracy, Adaboost has the best error bound. On th...
متن کاملPredicting Musical Eras of Songs
The goal of the project is to predict the year in which a certain piece of music was created. We used a subset of Million Song Dataset written for standardized tests to train different models including Naive Bayes Classifier, Generalized Linear Model, Random Forest, and Gradient Boosting Machine. Comparisons are made based on the results(mean square error and feature importance) from various tr...
متن کاملDiagnosis of Pulmonary Tuberculosis Using Artificial Intelligence (Naive Bayes Algorithm)
Background and Aim: Despite the implementation of effective preventive and therapeutic programs, no significant success has been achieved in the reduction of tuberculosis. One of the reasons is the delay in diagnosis. Therefore, the creation of a diagnostic aid system can help to diagnose early Tuberculosis. The purpose of this research was to evaluate the role of the Naive Bayes algorithm as a...
متن کاملLazy Bayesian Rules: A Lazy Semi-Naive Bayesian Learning Technique Competitive to Boosting Decision Trees
Lbr is a lazy semi-naive Bayesian classiier learning technique, designed to alleviate the attribute interdependence problem of naive Bayesian classiication. To classify a test example , it creates a conjunctive rule that selects a most appropriate subset of training examples and induces a local naive Bayesian classiier using this subset. Lbr can signii-cantly improve the performance of the naiv...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2009